Preliminary Experiments on Crowdsourced Evaluation of Feedback Granularity
نویسندگان
چکیده
Providing writing feedback to English language learners (ELLs) helps them learn to write better, but it is not clear what type or how much information should be provided. There have been few experiments directly comparing the effects of different types of automatically generated feedback on ELL writing. Such studies are difficult to conduct because they require participation and commitment from actual students and their teachers, over extended periods of time, and in real classroom settings. In order to avoid such difficulties, we instead conduct a crowdsourced study on Amazon Mechanical Turk to answer questions concerning the effects of type and amount of writing feedback. We find that our experiment has several serious limitations but still yields some interesting results.
منابع مشابه
Web-Based VR Experiments Powered by the Crowd
We build on the increasing availability of Virtual Reality (VR) devices and Web technologies to conduct behavioral experiments in VR using crowdsourcing techniques. A new recruiting and validation method allows us to create a panel of eligible experiment participants recruited from Amazon Mechanical Turk. Using this panel, we ran three different crowdsourced VR experiments, each reproducing one...
متن کاملCrowdsourced Top-k Algorithms: An Experimental Evaluation
Crowdsourced top-k computation has attracted significant attention recently, thanks to emerging crowdsourcing platforms, e.g., Amazon Mechanical Turk and CrowdFlower. Crowdsourced top-k algorithms ask the crowd to compare the objects and infer the top-k objects based on the crowdsourced comparison results. The crowd may return incorrect answers, but traditional top-k algorithms cannot tolerate ...
متن کاملCrowd-Selection Query Processing in Crowdsourcing Databases: A Task-Driven Approach
Crowd-selection is essential to crowdsourcing applications, since choosing the right workers with particular expertise to carry out specific crowdsourced tasks is extremely important. The central problem is simple but tricky: given a crowdsourced task, who is the right worker to ask? Currently, most existing work has mainly studied the problem of crowd-selection for simple crowdsourced tasks su...
متن کاملProfessionalism Mini-Evaluation Exercise in Finland: a preliminary investigation introducing the Finnish version of the P-MEX instrument
Introduction: Teaching medical professionalism is increasinglyacknowledged as an important aspect of medical education.The Professionalism Mini-Evaluation Exercise (P-MEX) is anassessment tool for evaluating medical professionalism, butno studies using it as a self-assessment instrument have beenreported. This paper reports on a preliminary investigation of theFinnish version of the P-MEX instr...
متن کاملTraining a Super Model Look-Alike: Featuring Edit Distance, N-Gram Occurrence, and One Reference Translation
Two string comparison measures, edit distance and n-gram co-occurrence, are tested for automatic evaluation of translation quality, where the quality is compared to one or several reference translations. The measures are tested in combination for diagnostic evaluation on segments. Both measures have been used for evaluation of translation quality before, but for another evaluation purpose (perf...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2015